where λ is the eigenvalue corresponding to the eigenvector x.
Given an eigenvalue λ, we can rearrange the above equation and solve for its corresponding eigenvector as a solution to the homogeneous system:
Ax−λx=0(A−λI)x=0
Eigenspace
The set of all solutions x to (A−λiI)x=0 is called the eigenspace of A corresponding to an eigenvalue λi, which includes 0 and all the eigenvectors x corresponding to λi.
Theorem 1
The eigenvalues of a triangular matrix are the entries on its main diagonal.
Theorem 2
If v1,…,vr are eigenvectors that correspond to distinct eigenvalues λ1,…,λr of an n×n matrix A, then the set {v1,…,vr} is linearly independent.
Intuitively, this makes sense because we can’t have two different stretching factors λ1 and λ2 along the same eigenvector x.
Questions
Is λ=3 and eigenvalue of A=1302−21211? If so, find one corresponding eigenvector.
Starting with the definition of an eigenvector:
Ax=λx
We rearrange the equation
(A−λI)x=0
Now, λ=3 is an eigenvalue only if the following system has a non-zero solution x:
(A−3I)x=0
We find a solution by row-reducing the matrix A−3I
A−3I=1−3302−2−31211−3∼100010001
This system has only the trivial solution x=0, which is not a valid eigenvector by definition. Thus, 3 is not an eigenvalue of A.
Find a basis for the eigenspace corresponding to λ=−2,5 of the matrix A=[1342]
The eigenspace is the set of all solutions x to the homogeneous equation (A−λiI)x=0. Using the given eigenvalues:
(A−(−2)I)x=0[3344]x=0→x=[−11]
So the eigenspace corresponding to λ=−2 is all vectors along [−11]
Similarly, for λ=5
(A−5I)x=0[−434−3]x=0→x=[11]
True or False: To find the eigenvalues of A, reduce A to echelon form.
False. Row operations change the characteristic polynomial, thus changing the eigenvalues. We find eigenvalues by solving for the roots of the characteristic polynomial det(A−λI)=0
5.2 The Characteristic Equation
The eigenvalues of matrix A are all scalars λi such that the homogeneous system
(A−λI)x=0
has a nontrivial solution x=0. This problem is equivalent to finding all λi such that the matrix A−λI is not invertible (by Invertible Matrix Theorem the homogeneous system only has a nontrivial solution when the matrix is not invertible). We know that a matrix is not invertible iff:
det(A−λI)=0
The above equation is called the characteristic equation. The algebraic multiplicity of an eigenvalue λi is its multiplicity as a root of the characteristic equation.
Theorem 4
If n×n matrices A and B are similar, then they have the same characteristic polynomial and hence the same eigenvalues with the same multiplicities.
Note: Similarity is not the same as row equivalence.
Questions
Find the characteristic polynomial and eigenvalues of A=[72−23]
det(A−λI)=07−λ2−23−λ=0(7−λ)(3−λ)+4=0
Given the above characteristic polynomial, we solve for eigenvalues by finding the roots.
λ2−10λ+25=0(λ−5)2=0
We find that A has the eigenvalue λ=5 with algebraic multiplicity (exponent) 2.
List the eigenvalues of Awith their multiplicities.
A=58010−47−500120001
By Theorem 1, we know that the eigenvalues of a triangular matrix are the entries on its diagonal. Thus, λ=1,1,−4,5.
Find h in the matrix A such that the eigenspace of λ=5 is 2-dimensional.
A=5000−23006h50−1041
The eigenspace of an eigenvalue is the set of all solutions x to (A−λiI)x=0.
0000−2−2006h00−104−4x1x2x3x4=0
∼00001000−3h−6000010x1x2x3x4=0
If h=6, then this homogeneous system has 2 free variables, thus making it 2-dimensional.
Use a property of determinants to show that A and A⊤ have the same characteristic polynomial.
Let B=A−λI:
detB=detB⊤det(A−λI)=det(A⊤−λI)
5.3 Diagonalization
Theorem 5: The Diagonalization Theorem
An n×n matrix A is diagonalizable iff A has n linearly independent eigenvectors.
A=PDP−1
where the n eigenvectors form the P matrix, and D is a diagonal matrix with the eigenvalues of A.
Example: Diagonalize A=1−333−533−31
First, we find the eigenvalues of A as the roots of the characteristic polynomial:
Then we find the eigenvalues corresponding to the two eigenvalues by plugging them into the homogeneous equation (A−λI)x=0 where the nonzero solutions x to this homogeneous system are eigenvectors corresponding to the eigenvalue. By the Diagonalization Theorem, there must be exactly n linearly independent eigenvectors for a matrix to be diagonalizable.
λ=1:1−11λ=−2−110,−101
Now we construct P and D:
P=1−11−110−101D=1000−2000−2
Theorem 6
An n×n matrix with n distinct eigenvalues is diagonalizable because it guarantees n linearly independent eigenvectors. This is a sufficient condition for a matrix to be diagonalizable. However, it doesn’t mean that a diagonalizable matrix must have n distinct eigenvalues.
Questions
Diagonalize A=[2431]
We find diagonal matrix D such that its entries are the two eigenvalues, and a matrix P such that its 2 columns are the 2 linearly independent eigenvectors corresponding to the eigenvalues.
We find the eigenvalues as roots to the characteristic polynomial:
The matrix A has 2 distinct eigenvalues, which is sufficient for diagonalization by Theorem 6 because it guarantees 2 linearly independent eigenvectors which guarantees that the matrix is diagonalizable by the Diagonalization Theorem.
D=[500−2]
We find the two eigenvectors as solutions to the homogeneous equations of their corresponding eigenvalues
We’re given the eigenvalues but not their multiplicities which is a bit annoying but it’s ok. Let’s just find the eigenvectors as the solution to the homogeneous system
We have 3 linearly independent eigenvectors, thus the matrix is diagonalizable by the Diagonalization Theorem:
P=−210−301−2−11D=200020001
True or False: If A is diagonalizable, then A is invertible.
False. A diagonalizable matrix may have λ=0 as an eigenvalue, which would make it non-invertible.
5.4 Eigenvectors and Linear Transformations
Theorem 8: Diagonal Matrix Represenation
Suppose A is diagonalizble as A=PDP−1. The columns of P form a basis for Rn, thus D is the P-coordinate matrix [T]P for the transformation x↦Ax.
Informally, this theorem says that a matrix diagonalization is expressing the matrix as a change of coordinates into the eigenspace, scaling along the axes (eigenvectors), and changing back into the original coordinate space. Where D expresses the transformation T in the eigenbasis P.
Thus, we can generalize T(x)=Ax to non-standard basis B:
[T(x)]B=[T]B[x]B
Proof
We have the eigenbasis P=[b1…bn]. We can show that the matrix D is the transformation T in the eigenbasis P:
[T]P=D
First, we remember that, in the standard basis, the columns ai of a linear transformation matrix A are defined as T(ei):
A=[T(e1)…T(en)]
We can then extend this definition beyond the standard basis, to the eigenbasis P:
[T]P=[[T(b1)]P…[T(bn)]P]
Then, again using the definition T(x)=Ax:
[T]P=[[Ab1]P…[Abn]P]
The change-of-coordinates matrix PE←P is just P, so PP←E=P−1
[T]P=[P−1Ab1…P−1Abn]
Factoring out both matrices:
[T]P=P−1A[b1…bn][T]P=P−1AP=D
The Matrix of a Linear Transformation
Given the coefficients [x]B of the basis we can find the coefficients of the transformed T(x) as :
[T(x)]B:
[T(x)]B=A[x]B
where A contains the transformed basis vectors
A=[[T(b1)]B…[T(bn)]B]
This is just a generalization of A=[T(e1)…T(en)] in the standard basis. Notice that A is not a change-of-coordinates matrix because both sides are coefficients of B-basis vectors.
Note: The set of all matrices similar to matrix A is equivalent to the set of all matrix representations of the transformation x↦Ax.
Questions
Let T:P2→P2 by T(p)=p(0)−p(1)t+p(2)t2.
a. Show that T is a linear transformation
First, let’s understand the problem. This linear transformation takes in a polynomial definition with a parameter t, for example p(t)=t. Then, we can plug this polynomial into the transformation as
Thus, we find that p(t)=−2+t is an eigenvector of this linear transformation with an eigenvalue of 1.
c. Find the matrix for T relative to the basis {1,t,t2} for P2
We know that the transformation matrix of a linear transformation is given by the transformed standard basis vectors:
M=[T(e1)…T(en)]
Thus, using the standard basis vectors in P2:
100=1,010=t,001=t2
T(1)=1−1∗t+1t2T(t)=0−1t+2t2T(t2)=0−1t+4t2
M=1−110−120−14
Let B={b1,b2,b3} be a basis for a vector space V. Find T(2b1−b2+4b3) when T is a linear transformation from V to V whose matrix relative to B is
[T]B=001−65−21−17
By Theorem 1, [T]B is just the generalization of the transformation matrix A to non-standard basis B such that